41 research outputs found

    Sketching for Large-Scale Learning of Mixture Models

    Get PDF
    Learning parameters from voluminous data can be prohibitive in terms of memory and computational requirements. We propose a "compressive learning" framework where we estimate model parameters from a sketch of the training data. This sketch is a collection of generalized moments of the underlying probability distribution of the data. It can be computed in a single pass on the training set, and is easily computable on streams or distributed datasets. The proposed framework shares similarities with compressive sensing, which aims at drastically reducing the dimension of high-dimensional signals while preserving the ability to reconstruct them. To perform the estimation task, we derive an iterative algorithm analogous to sparse reconstruction algorithms in the context of linear inverse problems. We exemplify our framework with the compressive estimation of a Gaussian Mixture Model (GMM), providing heuristics on the choice of the sketching procedure and theoretical guarantees of reconstruction. We experimentally show on synthetic data that the proposed algorithm yields results comparable to the classical Expectation-Maximization (EM) technique while requiring significantly less memory and fewer computations when the number of database elements is large. We further demonstrate the potential of the approach on real large-scale data (over 10 8 training samples) for the task of model-based speaker verification. Finally, we draw some connections between the proposed framework and approximate Hilbert space embedding of probability distributions using random features. We show that the proposed sketching operator can be seen as an innovative method to design translation-invariant kernels adapted to the analysis of GMMs. We also use this theoretical framework to derive information preservation guarantees, in the spirit of infinite-dimensional compressive sensing

    Multimodal Kalman Filtering

    No full text
    International audienceA difficult aspect of multimodal estimation is the possible discrepancybetween the sampling rates and/or the noise levels of theconsidered data. Many algorithms cope with these dissimilaritiesempirically. In this paper, we propose a conceptual analysis ofmultimodality where we try to find the “optimal” way of combiningmodalities. More specifically, we consider a simple Kalman filteringframework where several noisy sensors with different samplingfrequences and noise variances regularly observe a hidden state.We experimentally underline some relationships between the samplinggrids and the asymptotic variance of the maximum a posteriori(MAP) estimator. However, the explicit study of the asymptoticvariance seems intractable even in the simplest cases. We describe apromising idea to circumvent this difficulty: exploiting a stochasticmeasurement model for which one can more easily study the averageasymptotic behavior

    Fundamental performance limits for ideal decoders in high-dimensional linear inverse problems

    Get PDF
    This paper focuses on characterizing the fundamental performance limits that can be expected from an ideal decoder given a general model, ie, a general subset of "simple" vectors of interest. First, we extend the so-called notion of instance optimality of a decoder to settings where one only wishes to reconstruct some part of the original high dimensional vector from a low-dimensional observation. This covers practical settings such as medical imaging of a region of interest, or audio source separation when one is only interested in estimating the contribution of a specific instrument to a musical recording. We define instance optimality relatively to a model much beyond the traditional framework of sparse recovery, and characterize the existence of an instance optimal decoder in terms of joint properties of the model and the considered linear operator. Noiseless and noise-robust settings are both considered. We show somewhat surprisingly that the existence of noise-aware instance optimal decoders for all noise levels implies the existence of a noise-blind decoder. A consequence of our results is that for models that are rich enough to contain an orthonormal basis, the existence of an L2/L2 instance optimal decoder is only possible when the linear operator is not substantially dimension-reducing. This covers well-known cases (sparse vectors, low-rank matrices) as well as a number of seemingly new situations (structured sparsity and sparse inverse covariance matrices for instance). We exhibit an operator-dependent norm which, under a model-specific generalization of the Restricted Isometry Property (RIP), always yields a feasible instance optimality property. This norm can be upper bounded by an atomic norm relative to the considered model.Comment: To appear in IEEE Transactions on Information Theor

    Échantillonnage compressĂ© et rĂ©duction de dimension pour l'apprentissage non supervisĂ©

    No full text
    Cette thĂšse est motivĂ©e par la perspective de rapprochement entre traitement du signal et apprentissage statistique, et plus particuliĂšrement par l'exploitation de techniques d'Ă©chantillonnage compressĂ© afin de rĂ©duire le coĂ»t de tĂąches d'apprentissage. AprĂšs avoir rappelĂ© les bases de l'Ă©chantillonnage compressĂ© et mentionnĂ© quelques techniques d'analyse de donnĂ©es s'appuyant sur des idĂ©es similaires, nous proposons un cadre de travail pour l'estimation de paramĂštres de mĂ©lange de densitĂ©s de probabilitĂ© dans lequel les donnĂ©es d'entraĂźnement sont compressĂ©es en une reprĂ©sentation de taille fixe. Nous instancions ce cadre sur un modĂšle de mĂ©lange de Gaussiennes isotropes. Cette preuve de concept suggĂšre l'existence de garanties thĂ©oriques de reconstruction d'un signal pour des modĂšles allant au-delĂ  du modĂšle parcimonieux usuel de vecteurs. Nous Ă©tudions ainsi dans un second temps la gĂ©nĂ©ralisation de rĂ©sultats de stabilitĂ© de problĂšmes inverses linĂ©aires Ă  des modĂšles tout Ă  fait gĂ©nĂ©raux de signaux. Nous proposons des conditions sous lesquelles des garanties de reconstruction peuvent ĂȘtre donnĂ©es dans un cadre gĂ©nĂ©ral. Enfin, nous nous penchons sur un problĂšme de recherche approchĂ©e de plus proche voisin avec calcul de signature des vecteurs afin de rĂ©duire la complexitĂ©. Dans le cadre oĂč la distance d'intĂ©rĂȘt dĂ©rive d'un noyau de Mercer, nous proposons de combiner un plongement explicite des donnĂ©es suivi d'un calcul de signatures, ce qui aboutit notamment Ă  une recherche approchĂ©e plus prĂ©cise.This thesis is motivated by the perspective of connecting compressed sensing and machine learning, and more particularly by the exploitation of compressed sensing techniques to reduce the cost of learning tasks. After a reminder of compressed sensing and a quick description of data analysis techniques in which similar ideas are exploited, we propose a framework for estimating probability density mixture parameters in which the training data is compressed into a fixed-size representation. We instantiate this framework on an isotropic Gaussian mixture model. This proof of concept suggests the existence of theoretical guarantees for reconstructing signals belonging to models beyond usual sparse models. We therefore study generalizations of stability results for linear inverse problems for very general models of signals. We propose conditions under which reconstruction guarantees can be given in a general framework. Finally, we consider an approximate nearest neighbor search problem exploiting signatures of the database vectors in order to save resources during the search step. In the case where the considered distance derives from a Mercer kernel, we propose to combine an explicit embedding of data followed by a signature computation step, which principally leads to a more accurate approximate search

    Compressed sensing and dimensionality reduction for unsupervised learning

    No full text
    This thesis is motivated by the perspective of connecting compressed sensing and machine learning, and more particularly by the exploitation of compressed sensing techniques to reduce the cost of learning tasks. After a reminder of compressed sensing and a quick description of data analysis techniques in which similar ideas are exploited, we propose a framework for estimating probability density mixture parameters in which the training data is compressed into a fixed-size representation. We instantiate this framework on an isotropic Gaussian mixture model. This proof of concept suggests the existence of theoretical guarantees for reconstructing signals belonging to models beyond usual sparse models. We therefore study generalizations of stability results for linear inverse problems for very general models of signals. We propose conditions under which reconstruction guarantees can be given in a general framework. Finally, we consider an approximate nearest neighbor search problem exploiting signatures of the database vectors in order to save resources during the search step. In the case where the considered distance derives from a Mercer kernel, we propose to combine an explicit embedding of data followed by a signature computation step, which principally leads to a more accurate approximate search.Cette thĂšse est motivĂ©e par la perspective de rapprochement entre traitement du signal et apprentissage statistique, et plus particuliĂšrement par l'exploitation de techniques d'Ă©chantillonnage compressĂ© afin de rĂ©duire le coĂ»t de tĂąches d'apprentissage. AprĂšs avoir rappelĂ© les bases de l'Ă©chantillonnage compressĂ© et mentionnĂ© quelques techniques d'analyse de donnĂ©es s'appuyant sur des idĂ©es similaires, nous proposons un cadre de travail pour l'estimation de paramĂštres de mĂ©lange de densitĂ©s de probabilitĂ© dans lequel les donnĂ©es d'entraĂźnement sont compressĂ©es en une reprĂ©sentation de taille fixe. Nous instancions ce cadre sur un modĂšle de mĂ©lange de Gaussiennes isotropes. Cette preuve de concept suggĂšre l'existence de garanties thĂ©oriques de reconstruction d'un signal pour des modĂšles allant au-delĂ  du modĂšle parcimonieux usuel de vecteurs. Nous Ă©tudions ainsi dans un second temps la gĂ©nĂ©ralisation de rĂ©sultats de stabilitĂ© de problĂšmes inverses linĂ©aires Ă  des modĂšles tout Ă  fait gĂ©nĂ©raux de signaux. Nous proposons des conditions sous lesquelles des garanties de reconstruction peuvent ĂȘtre donnĂ©es dans un cadre gĂ©nĂ©ral. Enfin, nous nous penchons sur un problĂšme de recherche approchĂ©e de plus proche voisin avec calcul de signature des vecteurs afin de rĂ©duire la complexitĂ©. Dans le cadre oĂč la distance d'intĂ©rĂȘt dĂ©rive d'un noyau de Mercer, nous proposons de combiner un plongement explicite des donnĂ©es suivi d'un calcul de signatures, ce qui aboutit notamment Ă  une recherche approchĂ©e plus prĂ©cise

    Estimation de mélange de Gaussiennes sur données compressées

    Get PDF
    National audienceEstimating a probability mixture model from a set of vectors typically requires a large amount of memory if the data is voluminous. We propose a framework where the data is jointly compressed to a fixed-size representation called sketch, composed of empirical moments calculated from the data. By analogy with compressive sensing, we derive a parameter estimation algorithm from the sketch. We experimentally show that our algorithm allows precise estimation while consuming less memory than an EM algorithm for voluminous data. The algorithm also provides a privacy-preserving estimation tool since the sketch does not disclose information about individual datum it is based on

    Compressive Gaussian Mixture Estimation

    Get PDF
    International audienceWe propose a framework to estimate the parameters of a mixture of isotropic Gaussians using empirical data drawn from this mixture. The difference with standard methods is that we only use a sketch computed from the data instead of the data itself. The sketch is composed of empirical moments computed in one pass on the data. To estimate the mixture parameters from the sketch, we derive an algorithm by analogy with Iterative Hard Thresholding, used in compressed sensing to recover sparse signals from a few linear projections. We prove experimentally that the parameters can be precisely estimated if the sketch is large enough, while using less memory than an EM algorithm if the data is numerous. Our approach also preserves the privacy of the initial data, since the sketch doesn't provide information on the individual data

    Compressive Gaussian Mixture Estimation

    Get PDF
    International audienceWhen performing a learning task on voluminous data, memory and computational time can become prohibitive. In this chapter, we propose a framework aimed at estimating the parameters of a density mixture on training data in a compressive manner by computing a low-dimensional sketch of the data. The sketch represents empirical moments of the underlying probability distribution. Instantiating the framework on the case where the densities are isotropic Gaussians, we derive a reconstruction algorithm by analogy with compressed sensing. We experimentally show that it is possible to precisely estimate the mixture parameters provided that the sketch is large enough, while consuming less memory in the case of numerous data. The considered framework also provides a privacy-preserving data analysis tool, since the sketch does not disclose information about individual datum it is based on
    corecore